Mixup Without Hesitation

نویسندگان

چکیده

Mixup linearly interpolates pairs of examples to form new samples, which has been shown be effective in image classification tasks. However, there are two drawbacks mixup: one is that more training epochs needed obtain a well-trained model; the other mixup requires tuning hyper-parameter gain appropriate capacity. In this paper, we find constantly explores representation space, and inspired by exploration-exploitation dilemma, propose Without hesitation (mWh), concise algorithm. We show mWh strikes good balance between exploration exploitation gradually replacing with basic data augmentation. It can achieve strong baseline less time than original without searching for optimal hyper-parameter, i.e., acts as hesitation.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

mixup: Beyond Empirical Risk Minimization

Large deep neural networks are powerful, but exhibit undesirable behaviors such as memorization and sensitivity to adversarial examples. In this work, we propose mixup, a simple learning principle to alleviate these issues. In essence, mixup trains a neural network on convex combinations of pairs of examples and their labels. By doing so, mixup regularizes the neural network to favor simple lin...

متن کامل

mixup: BEYOND EMPIRICAL RISK MINIMIZATION

Large deep neural networks are powerful, but exhibit undesirable behaviors such as memorization and sensitivity to adversarial examples. In this work, we propose mixup, a simple learning principle to alleviate these issues. In essence, mixup trains a neural network on convex combinations of pairs of examples and their labels. By doing so, mixup regularizes the neural network to favor simple lin...

متن کامل

mixup: BEYOND EMPIRICAL RISK MINIMIZATION

Large deep neural networks are powerful, but exhibit undesirable behaviors such as memorization and sensitivity to adversarial examples. In this work, we propose mixup, a simple learning principle to alleviate these issues. In essence, mixup trains a neural network on convex combinations of pairs of examples and their labels. By doing so, mixup regularizes the neural network to favor simple lin...

متن کامل

Prosodic Cues for Hesitation

In our efforts to model spontaneous speech for use in, for example, spoken dialogue systems, a series of experiments have been conducted in order to investigate correlates to perceived hesitation. Previous work has shown that it is the total duration increase that is the valid cue rather than the contribution by either of the two factors pause duration and final lengthening. In the present expe...

متن کامل

The Virtues of Hesitation‡

In many economic, political and social situations, circumstances change at random points in time, reacting is costly, and changes appropriate to present circumstances may be inappropriate after later changes, requiring further costly change. Waiting is informative if the hazard rate for the arrival of the next change is non-constant. In a broad range of contexts we show that hesitation or delay...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Lecture Notes in Computer Science

سال: 2021

ISSN: ['1611-3349', '0302-9743']

DOI: https://doi.org/10.1007/978-3-030-87358-5_12